Cross-Layer Distillation with Semantic Calibration

نویسندگان

چکیده

Recently proposed knowledge distillation approaches based on feature-map transfer validate that intermediate layers of a teacher model can serve as effective targets for training student to obtain better generalization ability. Existing studies mainly focus particular representation forms between manually specified pairs teacher-student layers. However, semantics may vary in different networks and manual association might lead negative regularization caused by semantic mismatch certain layer pairs. To address this problem, we propose Semantic Calibration Cross-layer Knowledge Distillation (SemCKD), which automatically assigns proper target the each with an attention mechanism. With learned distribution, distills contained multiple rather than single fixed from appropriate cross-layer supervision training. Consistent improvements over state-of-the-art are observed extensive experiments various network architectures models, demonstrating effectiveness flexibility soft mechanism distillation.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Spreadsheets with a Semantic Layer

Spreadsheets are active documents that are heavily employed in administration, financial forecasting, education, and science because of their intuitive, flexible, and direct approach to computation. But they are also error-prone, poorly documented, often contain actual data in legacy form. Therefore, assistance for high-impact spreadsheet users is needed. To determine what kind of help could be...

متن کامل

Term Distillation for Cross-DB Retrieval

In cross-DB retrieval, the domain of queries differs from the retrieval target in the distribution of that of term occurrences. This causes incorrect term weighting in the retrieval system which assigns to each term a retrieval weight based on the distribution of term occurrences. To resolve the problem, we propose \term distillation" which is a framework for query term selection in crossDB ret...

متن کامل

Iraia: a Portal Technology with a Semantic Layer Coordinating Multimedia Retrieval and Cross-owner Content Building

This paper presents a technology for information portals that supports multimedia retrieval while focusing at the same time on different ecologies of information provider and content owner environments. This means it addresses cross-media information provision and cross-owner content building. The technology was developed under the FP5 project IRAIA and was applied to two different areas: econo...

متن کامل

Cross-lingual Distillation for Text Classification

Cross-lingual text classification(CLTC) is the task of classifying documents written in different languages into the same taxonomy of categories. This paper presents a novel approach to CLTC that builds on model distillation, which adapts and extends a framework originally proposed for model compression. Using soft probabilistic predictions for the documents in a label-rich language as the (ind...

متن کامل

Cross-lingual sentence extraction for information distillation

Information distillation aims to analyze and interpret large volumes of speech and text archives in multiple languages and produce structured information of interest to the user. In this work, we investigate cross-lingual information distillation, where nonEnglish (source language) documents are searched for user queries that are in English (target language). We propose to perform distillation ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i8.16865